Successive approximation methods for Markov games

نویسندگان

  • J. van der Wal
  • J. Wessels
چکیده

• A submitted manuscript is the author's version of the article upon submission and before peer-review. There can be important differences between the submitted version and the official published version of record. People interested in the research are advised to contact the author for the final version of the publication, or visit the DOI to the publisher's website. • The final author version and the galley proof are versions of the publication after peer review. • The final published version features the final layout of the paper including the volume, issue and page numbers.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A constructive study of Markov equilibria in stochastic games with strategic complementarities

We study a class of discounted infinite horizon stochastic games with strategic complementarities. Using monotone operators on the space of values and strategies, we prove existence of a Stationary Markov Nash equilibrium under different set of assumptions than Curtat (1996), Amir (2002, 2005) or Nowak (2007) via constructive methods. In addition, we provide monotone comparative statics results...

متن کامل

A Novel Successive Approximation Method for Solving a Class of Optimal Control Problems

This paper presents a successive approximation method (SAM) for solving a large class of optimal control problems. The proposed analytical-approximate method, successively solves the Two-Point Boundary Value Problem (TPBVP), obtained from the Pontryagin's Maximum Principle (PMP). The convergence of this method is proved and a control design algorithm with low computational complexity is present...

متن کامل

Improved successive approximation methods for discounted Markov decision processes

Successive Approximation (S.A.) methods, for solving discounted Markov decision problems, have been developed to avoid the extensive computations that are connected with linear programming and policy iteration techniques for solving large scaled problems. Several authors give such an S.A. algorithm. In this paper we introduce some new algorithms while furthermore it will be shown how the severa...

متن کامل

Value Function Approximation in Zero-Sum Markov Games

This paper investigates value function approximation in the context of zero-sum Markov games, which can be viewed as a generalization of the Markov decision process (MDP) framework to the two-agent case. We generalize error bounds from MDPs to Markov games and describe generalizations of reinforcement learning algorithms to Markov games. We present a generalization of the optimal stopping probl...

متن کامل

Time consistent Markov policies in dynamic economies with quasi-hyperbolic consumers

We study the question of existence and computation of time-consistent Markov policies of quasi-hyperbolic consumers under a stochastic transition technology in a general class of economies with multidimensional action spaces and uncountable state spaces. Under standard complementarity assumptions on preferences, as well as a mild geometric condition on a transition probabilities, we prove exist...

متن کامل

Successive approximations for Markov decision processes and Markov games with unbounded rewards

• A submitted manuscript is the author's version of the article upon submission and before peer-review. There can be important differences between the submitted version and the official published version of record. People interested in the research are advised to contact the author for the final version of the publication, or visit the DOI to the publisher's website. • The final author version ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2017